AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Multi-query attention

# Multi-query attention

Bert 1.3b
Apache-2.0
Transformer encoder pretrained based on Megatron-LM, specifically designed for Japanese scenarios
Large Language Model Transformers Supports Multiple Languages
B
retrieva-jp
56
15
Falcon 7B Instruct GPTQ
Apache-2.0
The 4-bit quantized version of Falcon-7B-Instruct, quantized using the AutoGPTQ tool, suitable for efficient inference in resource-constrained environments.
Large Language Model Transformers English
F
TheBloke
189
67
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase